Finding the Value of Information About a State Variable in a Markov Decision Process

نویسندگان

  • Gilvan C. Souza
  • Robert H. Smith
چکیده

Finding the Value of Information About a State Variable in a Markov Decision Process Gilvan C. Souza The Robert H. Smith School of Business, The University of Maryland, College Park, MD, 20742 Phone: 301-405-0628; fax: 301-405-8655; email: [email protected] ABSTRACT In this paper we present a mixed–integer programming formulation that computes the optimal solution for a certain class of Markov decision processes with finite state and action spaces, where a state is comprised of multiple state variables, and one of the state variables is unobservable to the decision maker. Our approach is a much simpler modeling alternative to the theory of partially observable Markov decision processes (POMDP), where an information and updating structure about the decision variable needs to be defined. We illustrate the approach with an example of a duopoly where one firm’s actions are not immediately observable by the other firm, and present computational results. We believe that this approach can be used in a variety of applications, where the decision maker wants to assess the value of information about an additional decision variable.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Accelerated decomposition techniques for large discounted Markov decision processes

Many hierarchical techniques to solve large Markov decision processes (MDPs) are based on the partition of the state space into strongly connected components (SCCs) that can be classified into some levels. In each level, smaller problems named restricted MDPs are solved, and then these partial solutions are combined to obtain the global solution. In this paper, we first propose a novel algorith...

متن کامل

Utilizing Generalized Learning Automata for Finding Optimal Policies in MMDPs

Multi agent Markov decision processes (MMDPs), as the generalization of Markov decision processes to the multi agent case, have long been used for modeling multi agent system and are used as a suitable framework for Multi agent Reinforcement Learning. In this paper, a generalized learning automata based algorithm for finding optimal policies in MMDP is proposed. In the proposed algorithm, MMDP ...

متن کامل

Decisions of Value: Going Backstage; Comment on “Contextual Factors Influencing Cost and Quality Decisions in Health and Care: A Structured Evidence Review and Narrative Synthesis”

This commentary expands on two of the key themes briefly raised in the paper involving analysis of the evidence about key contextual influences on decisions of value. The first theme focuses on the need to explore in more detail what is called backstage decision-making looking at how actual decisions are made drawing on evidence from ethnographies about decision-making. These studies point to l...

متن کامل

Decision making in forest management with consideration of stochastic prices

  The optimal harvesting policy is calculated as a function of the entering stock, the price state, the harvesting cost, and the rate of interest in the capital market. In order to determine the optimal harvest schedule, the growth function and stumpage price process are estimated for the Swedish mixed species forests. The stumpage price is assumed to follow a stochastic Markov process. A stoch...

متن کامل

تبیین رابطه سیستمهای اطلاعاتی (IS) و فرآیند تصمیم گیری مدیران میانی نظام بانکداری خصوصی ( مورد مطالعه : بانک سامان )

In this thesis “the relationship between information systems (IS) and decision making process of private bank middle managers” on the aspect of decisions effectiveness enhancement is surveyed.The independent variable of this research is information systems that consists of TPS, MIS, DSS, GDSS and SIS. Decision making process could be affected from several aspects by information systems.Some imp...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2004